A Survey on Evaluation Metrics for Machine Translation
نویسندگان
چکیده
The success of Transformer architecture has seen increased interest in machine translation (MT). quality neural network-based MT transcends that translations derived using statistical methods. This growth research entailed the development accurate automatic evaluation metrics allow us to track performance MT. However, automatically evaluating and comparing systems is a challenging task. Several studies have shown traditional (e.g., BLEU, TER) show poor capturing semantic similarity between outputs human reference translations. To date, improve performance, various been proposed architecture. systematic comprehensive literature review on these still missing. Therefore, it necessary survey existing enable both established new researchers quickly understand trend over past few years. In this survey, we present metrics. better developments field, provide taxonomy Then, explain key contributions shortcomings addition, select representative from taxonomy, conduct experiments analyze related problems. Finally, discuss limitation current metric through experimentation our suggestions for further
منابع مشابه
Evaluation Metrics for Knowledge-Based Machine Translation
A methodology is presented for component-based machine translation (MT) evaluation through causal error analysis to complement existing global evaluation methods. This methodology is particularly appropriate for knowledge-based machine translation (KBMT) systems. After a discussion of MT evaluation criteria and the particular evaluation metrics proposed for KBMT, we apply this methodology to a ...
متن کاملMachine Translation Evaluation: A Survey
This paper introduces the state-of-the-art MT evaluation survey that contains both manual and automatic evaluation methods. The traditional human evaluation criteria mainly include the intelligibility, fidelity, fluency, adequacy, comprehension, and informativeness. We classify the automatic evaluation methods into two categories, including lexical similarity and linguistic features application...
متن کاملA Poor Man’s Translation Memory Using Machine Translation Evaluation Metrics
We propose straightforward implementations of translation memory (TM) functionality for research purposes, using machine translation evaluation metrics as similarity functions. Experiments under various conditions demonstrate the effectiveness of the approach, but also highlight problems in evaluating the results using an MT evaluation methodology.
متن کاملThe Correlation of Machine Translation Evaluation Metrics with Human Judgement on Persian Language
Machine Translation Evaluation Metrics (MTEMs) are the central core of Machine Translation (MT) engines as they are developed based on frequent evaluation. Although MTEMs are widespread today, their validity and quality for many languages is still under question. The aim of this research study was to examine the validity and assess the quality of MTEMs from Lexical Similarity set on machine tra...
متن کاملA Dataset for Assessing Machine Translation Evaluation Metrics
We describe a dataset containing 16,000 translations produced by four machine translation systems and manually annotated for quality by professional translators. This dataset can be used in a range of tasks assessing machine translation evaluation metrics, from basic correlation analysis to training and test of machine learning-based metrics. By providing a standard dataset for such tasks, we h...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Mathematics
سال: 2023
ISSN: ['2227-7390']
DOI: https://doi.org/10.3390/math11041006